Myoelectric signals for multimodal speech recognition
نویسندگان
چکیده
A Coupled Hidden Markov Model (CHMM) is proposed in this paper to perform multimodal speech recognition using myoeletric signals (MES) from the muscles of vocal articulation. MES signals are immune to noise, and words that are acoustically similar manifest distinctly in MES. Hence, they would effectively complement the acoustic data in a multimodal speech recognition system. Research in Audio-Visual Speech Recognition has shown that CHMMs model the asynchrony between different data streams effectively. Hence, we propose CHMM for multimodal speech recognition using audio and MES as the two data streams. Our experiments indicate that the multimodal CHMM system significantly outperforms the audio only system at different SNRs. We have also provided a comparison between different features for MES and have found that wavelet features provide the best results.
منابع مشابه
A Comparative Study of Gender and Age Classification in Speech Signals
Accurate gender classification is useful in speech and speaker recognition as well as speech emotion classification, because a better performance has been reported when separate acoustic models are employed for males and females. Gender classification is also apparent in face recognition, video summarization, human-robot interaction, etc. Although gender classification is rather mature in a...
متن کاملUnspoken Words Recognition: A Review
In recent years, unspoken words recognition has received substantial attention from both the scientific research communities and the society of multimedia information access networks. Major advancements and wide range of applications in aids for the speech handicapped, speech pathology research, telecom privacy issues, cursor based text to speech, firefighters wearing pressurized suits with sel...
متن کاملP65: Speech Recognition Based on Bbrain Signals by the Quantum Support Vector Machine for Inflammatory Patient ALS
People communicate with each other by exchanging verbal and visual expressions. However, paralyzed patients with various neurological diseases such as amyotrophic lateral sclerosis and cerebral ischemia have difficulties in daily communications because they cannot control their body voluntarily. In this context, brain-computer interface (BCI) has been studied as a tool of communication for thes...
متن کاملRobust Methods for EMG Signal Processing for Audio-EMG- based Multi-modal Speech Recognition
This paper proposes robust methods for processing EMG (electromyography) signals in the framework of audio-EMGbased speech recognition. The EMG signals are captured when uttered and used as auxiliary information for recognizing speech. Two robust methods (Cepstral Mean Normalization and Spectral Subtraction) for EMG signal processing are investigated to improve the recognition performance. We a...
متن کاملSmart multimodal interfaces for human-computer interaction on train stations
This paper describes some approaches to processing audio and video signals with support for speech recognition. Speech recognition and lip-reading are currently very popular research topics in the scientific community. They deal with the audio and the video signals and their interactions with each other. The goal of this paper is to show how simple methods can be used for both modalities. Namel...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2005